What Do You MEME? Generating Explanations for Visual Semantic Role Labelling in Memes

نویسندگان

چکیده

Memes are powerful means for effective communication on social media. Their effortless amalgamation of viral visuals and compelling messages can have far-reaching implications with proper marketing. Previous research memes has primarily focused characterizing their affective spectrum detecting whether the meme's message insinuates any intended harm, such as hate, offense, racism, etc. However, often use abstraction, which be elusive. Here, we introduce a novel task - EXCLAIM, generating explanations visual semantic role labeling in memes. To this end, curate ExHVV, dataset that offers natural language connotative roles three types entities heroes, villains, victims, encompassing 4,680 present 3K We also benchmark ExHVV several strong unimodal multimodal baselines. Moreover, posit LUMEN, multimodal, multi-task learning framework endeavors to address EXCLAIM optimally by jointly predict correct correspondingly generate suitable explanations. LUMEN distinctly outperforms best baseline across 18 standard generation evaluation metrics. Our systematic analyses demonstrate characteristic cues required adjudicating helpful

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Visual memory: What do you know about what you saw?

Recent studies of visual perception are bringing us closer to an understanding of what we remember - and what we forget - when we recall a scene.

متن کامل

Generating Visual Explanations

Clearly explaining a rationale for a classification decision to an end user can be as important as the decision itself. Existing approaches for deep visual recognition are generally opaque and do not output any justification text; contemporary vision-language models can describe image content but fail to take into account class-discriminative image aspects which justify visual predictions. We p...

متن کامل

Do you do what you say or do you do what you say others do?

We design a donations vs. own money choice experiment comparing three different treatments. In two of the treatments the pay-offs are hypothetical. In the first of these, a short cheap talk script was used, and subjects were required to state their own preferences in this scenario. In the second, subjects were asked to state how they believed an average student would respond to the choices. In ...

متن کامل

Crowdsourced Explanations for Humorous Internet Memes

The subculture of humorous images is a pervasive phenomenon on the Internet. Some social media websites (e.g. 9GAG1, Reddit and 4chan) provide platforms for people to share their images and comments. These websites are often occupied by experienced users, and most of the newcomers have trouble fitting in such communities. Among all the types of humorous images, Internet memes are relatively har...

متن کامل

Visual transformations in gesture imitation: what you see is what you do

We propose an approach for a robot to imitate the gestures of a human demonstrator. Our framework consists solely of two components: a Sensory-Motor Map (SMM) and a View-Point Transformation (VPT). The SMM establishes an association between an arm image and the corresponding joint angles and it is learned by the system during a period of observation of its own gestures. The VPT is widely discus...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2023

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v37i8.26166